# Lightweight Vision-Language
Nanovlm 450M
MIT
nanoVLM is a lightweight vision-language model (VLM) designed for efficient training and experimentation.
Image-to-Text
Safetensors
N
lusxvr
339
2
Nanovlm
MIT
nanoVLM is a lightweight vision-language model (VLM) designed for efficient training and experimentation.
Image-to-Text
Safetensors
N
andito
187
1
Nanovlm 222M
Apache-2.0
nanoVLM is an ultra-minimalist lightweight vision-language model (VLM) designed for efficient training and experimentation.
Image-to-Text
N
lusxvr
2,441
73
Smolvlm2 2.2B Instruct GGUF
Apache-2.0
SmolVLM2-2.2B-Instruct is a 2.2B parameter vision-language model focused on video-text-to-text tasks, supporting English.
English
S
mradermacher
235
0
R1 VL 2B
Apache-2.0
R1-VL-2B is a vision-language reasoning model trained through Stepwise Group Relative Policy Optimization (StepGRPO), optimized based on Qwen2-VL-2B-Instruct.
Image-to-Text
Transformers

R
jingyiZ00
272
1
Moondream2
Apache-2.0
Moondream is a lightweight vision-language model designed for efficient operation across all platforms.
Image-to-Text
M
vikhyatk
184.93k
1,120
Featured Recommended AI Models